How GPT Actually Works
A detailed look at the entire process of how GPT takes sentences as input and generates responses. From tokenization to Transformer structure to parameter counts.
4 posts
A detailed look at the entire process of how GPT takes sentences as input and generates responses. From tokenization to Transformer structure to parameter counts.
Breaking down handwritten digit recognition code line by line to understand how Sequential, Dense, ReLU, Softmax, and training actually work.
A beginner-friendly guide covering AI, machine learning, deep learning, Transformer, and GPT - understanding the complete picture.
What is sparse_categorical_crossentropy commonly used in machine learning? Why do we need loss functions? Explained simply.